Goto

Collaborating Authors

 cnn 2




Viewpoint Generalization via a Binocular Vision

Neural Information Processing Systems

Convolutional Neural Networks (CNNs, LeCun et al. (1989, 1998)) are models inspired by how the Despite giving impressive performance in many applications, CNNs still have a long way to go in terms of being comparable to human's visual ability. The two images from the left eye and the right eye are merged and then fed into regular CNN layers.


dabd8d2ce74e782c65a973ef76fd540b-AuthorFeedback.pdf

Neural Information Processing Systems

Fig. (b) shows the performance of vanilla CNN with Fig. (c) shows that CNN The experimental settings are toy-like. ICLR'18, which considered only grayscale images. The neural network backbone is weak. The results are shown in Fig. (e). We hope our above explanation relieves your concerns, and if so, please consider raising your score.


Minimal Batch Adaptive Learning Policy Engine for Real-Time Mid-Price Forecasting in High-Frequency Trading

Ntakaris, Adamantios, Ibikunle, Gbenga

arXiv.org Artificial Intelligence

High-frequency trading (HFT) has transformed modern financial markets, making reliable short-term price forecasting models essential. In this study, we present a novel approach to mid-price forecasting using Level 1 limit order book (LOB) data from NASDAQ, focusing on 100 U.S. stocks from the S&P 500 index during the period from September to November 2022. Expanding on our previous work with Radial Basis Function Neural Networks (RBFNN), which leveraged automated feature importance techniques based on mean decrease impurity (MDI) and gradient descent (GD), we introduce the Adaptive Learning Policy Engine (ALPE) - a reinforcement learning (RL)-based agent designed for batch-free, immediate mid-price forecasting. ALPE incorporates adaptive epsilon decay to dynamically balance exploration and exploitation, outperforming a diverse range of highly effective machine learning (ML) and deep learning (DL) models in forecasting performance.


Testing Neural Network Verifiers: A Soundness Benchmark with Hidden Counterexamples

Zhou, Xingjian, Xu, Hongji, Xu, Andy, Shi, Zhouxing, Hsieh, Cho-Jui, Zhang, Huan

arXiv.org Artificial Intelligence

In recent years, many neural network (NN) verifiers have been developed to formally verify certain properties of neural networks such as robustness. Although many benchmarks have been constructed to evaluate the performance of NN verifiers, they typically lack a ground-truth for hard instances where no current verifier can verify and no counterexample can be found, which makes it difficult to check the soundness of a new verifier if it claims to verify hard instances which no other verifier can do. We propose to develop a soundness benchmark for NN verification. Our benchmark contains instances with deliberately inserted counterexamples while we also try to hide the counterexamples from regular adversarial attacks which can be used for finding counterexamples. We design a training method to produce neural networks with such hidden counterexamples. Our benchmark aims to be used for testing the soundness of NN verifiers and identifying falsely claimed verifiability when it is known that hidden counterexamples exist. We systematically construct our benchmark and generate instances across diverse model architectures, activation functions, input sizes, and perturbation radii. We demonstrate that our benchmark successfully identifies bugs in state-of-the-art NN verifiers, as well as synthetic bugs, providing a crucial step toward enhancing the reliability of testing NN verifiers.


Machine-Learning Kronecker Coefficients

Lee, Kyu-Hwan

arXiv.org Machine Learning

The Kronecker coefficients are the decomposition multiplicities of the tensor product of two irreducible representations of the symmetric group. Unlike the Littlewood--Richardson coefficients, which are the analogues for the general linear group, there is no known combinatorial description of the Kronecker coefficients, and it is an NP-hard problem to decide whether a given Kronecker coefficient is zero or not. In this paper, we show that standard machine-learning algorithms such as Nearest Neighbors, Convolutional Neural Networks and Gradient Boosting Decision Trees may be trained to predict whether a given Kronecker coefficient is zero or not. Our results show that a trained machine can efficiently perform this binary classification with high accuracy ($\approx 0.98$).


Bayesian optimization for sparse neural networks with trainable activation functions

Fakhfakh, Mohamed, Chaari, Lotfi

arXiv.org Artificial Intelligence

In the literature on deep neural networks, there is considerable interest in developing activation functions that can enhance neural network performance. In recent years, there has been renewed scientific interest in proposing activation functions that can be trained throughout the learning process, as they appear to improve network performance, especially by reducing overfitting. In this paper, we propose a trainable activation function whose parameters need to be estimated. A fully Bayesian model is developed to automatically estimate from the learning data both the model weights and activation function parameters. An MCMC-based optimization scheme is developed to build the inference. The proposed method aims to solve the aforementioned problems and improve convergence time by using an efficient sampling scheme that guarantees convergence to the global maximum. The proposed scheme is tested on three datasets with three different CNNs. Promising results demonstrate the usefulness of our proposed approach in improving model accuracy due to the proposed activation function and Bayesian estimation of the parameters.